Nyquist–Shannon sampling theorem

Fig.1: Hypothetical spectrum of a bandlimited signal as a function of frequency

The Nyquist–Shannon sampling theorem, named after Harry Nyquist and Claude Shannon, is a fundamental result in the field of information theory, in particular telecommunications and signal processing. Sampling is the process of converting a signal (for example, a function of continuous time or space) into a numeric sequence (a function of discrete time or space). Shannon's version of the theorem states:[1]

If a function x(t) contains no frequencies higher than B hertz, it is completely determined by giving its ordinates at a series of points spaced 1/(2B) seconds apart.

The theorem is commonly called the Nyquist sampling theorem, and is also known as Nyquist–Shannon–Kotelnikov, Whittaker–Shannon–Kotelnikov, Whittaker–Nyquist–Kotelnikov–Shannon, WKS, etc., sampling theorem, as well as the Cardinal Theorem of Interpolation Theory. It is often referred to as simply the sampling theorem.

In essence, the theorem shows that a bandlimited analog signal that has been sampled can be perfectly reconstructed from an infinite sequence of samples if the sampling rate exceeds 2B samples per second, where B is the highest frequency in the original signal. If a signal contains a component at exactly B hertz, then samples spaced at exactly 1/(2B) seconds do not completely determine the signal, Shannon's statement notwithstanding. This sufficient condition can be weakened, as discussed at Sampling of non-baseband signals below.

More recent statements of the theorem are sometimes careful to exclude the equality condition; that is, the condition is if x(t) contains no frequencies higher than or equal to B; this condition is equivalent to Shannon's except when the function includes a steady sinusoidal component at exactly frequency B.

The theorem assumes an idealization of any real-world situation, as it only applies to signals that are sampled for infinite time; any time-limited x(t) cannot be perfectly bandlimited. Perfect reconstruction is mathematically possible for the idealized model but only an approximation for real-world signals and sampling techniques, albeit in practice often a very good one.

The theorem also leads to a formula for reconstruction of the original signal. The constructive proof of the theorem leads to an understanding of the aliasing that can occur when a sampling system does not satisfy the conditions of the theorem.

The sampling theorem provides a sufficient condition, but not a necessary one, for perfect reconstruction. The field of compressed sensing provides a stricter sampling condition when the underlying signal is known to be sparse. Compressed sensing specifically yields a sub-Nyquist sampling criterion.

Contents

Introduction

A signal or function is bandlimited if it contains no energy at frequencies higher than some bandlimit or bandwidth B. A signal that is bandlimited is constrained in how rapidly it changes in time, and therefore how much detail it can convey in an interval of time. The sampling theorem asserts that the uniformly spaced discrete samples are a complete representation of the signal if this bandwidth is less than half the sampling rate. To formalize these concepts, let x(t) represent a continuous-time signal and X(f) be the continuous Fourier transform of that signal:

X(f)\ \stackrel{\mathrm{def}}{=}\ \int_{-\infty}^{\infty} x(t) \ e^{- i 2 \pi f t} \ dt. \

The signal x(t) is said to be bandlimited to a one-sided baseband bandwidth, B, if:

X(f) = 0 \quad   for all  |{f}| > B\,

or, equivalently, supp(X)[2]  \subseteq  [−BB].  Then the sufficient condition for exact reconstructability from samples at a uniform sampling rate fs (in samples per unit time) is:

f_s > 2 B,\,

or equivalently:

B < { f_s  \over 2 }.\,

2B is called the Nyquist rate and is a property of the bandlimited signal, while fs / 2 is called the Nyquist frequency and is a property of this sampling system.

The time interval between successive samples is referred to as the sampling interval:

T\ \stackrel{\mathrm{def}}{=}\ \frac{1}{f_s},\,

and the samples of x(t) are denoted by:

x[n] = x(nT) \quad n\in\mathbb{Z}\,(integers).

The sampling theorem leads to a procedure for reconstructing the original x(t) from the samples and states sufficient conditions for such a reconstruction to be exact.

The sampling process

The theorem describes two processes in signal processing: a sampling process, in which a continuous time signal is converted to a discrete time signal, and a reconstruction process, in which the original continuous signal is recovered from the discrete time signal.

The continuous signal varies over time (or space in a digitized image, or another independent variable in some other application) and the sampling process is performed by measuring the continuous signal's value every T units of time (or space), which is called the sampling interval. In practice, for signals that are a function of time, the sampling interval is typically quite small, on the order of milliseconds, microseconds, or less. This results in a sequence of numbers, called samples, to represent the original signal. Each sample value is associated with the instant in time when it was measured. The reciprocal of the sampling interval (1/T) is the sampling frequency denoted fs, which is measured in samples per unit of time. If T is expressed in seconds, then fs is expressed in Hz.

Reconstruction

Reconstruction of the original signal is an interpolation process that mathematically defines a continuous-time signal x(t) from the discrete samples x[n] and at times in between the sample instants nT.

Fig.2: The normalized sinc function: sin(πx) / (πx) ... showing the central peak at x= 0, and zero-crossings at the other integer values of x.

If the original signal contains a frequency component equal to one-half the sampling rate, the condition is not satisfied. The resulting reconstructed signal may have a component at that frequency, but the amplitude and phase of that component generally will not match the original component.

This reconstruction or interpolation using sinc functions is not the only interpolation scheme. Indeed, it is impossible in practice because it requires summing an infinite number of terms. However, it is the interpolation method that in theory exactly reconstructs any given bandlimited x(t) with any bandlimit B < 1/(2T); any other method that does so is formally equivalent to it.

Practical considerations

A few consequences can be drawn from the theorem:

Aliasing

The Poisson summation formula shows that the samples, x[n]=x(nT), of function x(t) are sufficient to create a periodic extension of function X(f). The result is:

X_s(f)\ \stackrel{\mathrm{def}}{=} \sum_{k=-\infty}^{\infty} X\left(f - k f_s\right) = T \sum_{n=-\infty}^{\infty} x(nT)\ e^{-i 2\pi n T f}.

 

 

 

 

(Eq.1)

As depicted in Figures 3, 4, and 8, copies of X(f) are shifted by multiples of fs and combined by addition.

Fig.3: Hypothetical spectrum of a properly sampled bandlimited signal (blue) and images (green) that do not overlap. A "brick-wall" low-pass filter can remove the images and leave the original spectrum, thus recovering the original signal from the samples.

If the sampling condition is not satisfied, adjacent copies overlap, and it is not possible in general to discern an unambiguous X(f). Any frequency component above fs/2 is indistinguishable from a lower-frequency component, called an alias, associated with one of the copies. The reconstruction technique described below produces the alias, rather than the original component, in such cases.

Fig.4 Top: Hypothetical spectrum of an insufficiently sampled bandlimited signal (blue), X(f), where the images (green) overlap. These overlapping edges or "tails" of the images add, creating a spectrum unlike the original. Bottom: Hypothetical spectrum of a marginally sufficiently sampled bandlimited signal (blue), XA(f), where the images (green) narrowly do not overlap. But the overall sampled spectrum of XA(f) is identical to the overall inadequately sampled spectrum of X(f) (top) because the sum of baseband and images are the same in both cases. The discrete sampled signals xA[n] and x[n] are also identical. It is not possible, just from examining the spectra (or the sampled signals), to tell the two situations apart. If this were an audio signal, xA[n] and x[n] would sound the same and the presumed "properly" sampled xA[n] would be the alias of x[n] since the spectrum XA(f) masquerades as the spectrum X(f).

For a sinusoidal component of exactly half the sampling frequency, the component will in general alias to another sinusoid of the same frequency, but with a different phase and amplitude.

To prevent or reduce aliasing, two things can be done:

  1. Increase the sampling rate, to above twice some or all of the frequencies that are aliasing.
  2. Introduce an anti-aliasing filter or make the anti-aliasing filter more stringent.

The anti-aliasing filter is to restrict the bandwidth of the signal to satisfy the condition for proper sampling. Such a restriction works in theory, but is not precisely satisfiable in reality, because realizable filters will always allow some leakage of high frequencies. However, the leakage energy can be made small enough so that the aliasing effects are negligible.

Application to multivariable signals and images

Fig.5: Subsampled image showing a Moiré pattern
Fig.6: See full size image

The sampling theorem is usually formulated for functions of a single variable. Consequently, the theorem is directly applicable to time-dependent signals and is normally formulated in that context. However, the sampling theorem can be extended in a straightforward way to functions of arbitrarily many variables. Grayscale images, for example, are often represented as two-dimensional arrays (or matrices) of real numbers representing the relative intensities of pixels (picture elements) located at the intersections of row and column sample locations. As a result, images require two independent variables, or indices, to specify each pixel uniquely — one for the row, and one for the column.

Color images typically consist of a composite of three separate grayscale images, one to represent each of the three primary colors — red, green, and blue, or RGB for short. Other colorspaces using 3-vectors for colors include HSV, LAB, XYZ, etc. Some colorspaces such as cyan, magenta, yellow, and black (CMYK) may represent color by four dimensions. All of these are treated as vector-valued functions over a two-dimensional sampled domain.

Similar to one-dimensional discrete-time signals, images can also suffer from aliasing if the sampling resolution, or pixel density, is inadequate. For example, a digital photograph of a striped shirt with high frequencies (in other words, the distance between the stripes is small), can cause aliasing of the shirt when it is sampled by the camera's image sensor. The aliasing appears as a moiré pattern. The "solution" to higher sampling in the spatial domain for this case would be to move closer to the shirt, use a higher resolution sensor, or to optically blur the image before acquiring it with the sensor.

Another example is shown to the left in the brick patterns. The top image shows the effects when the sampling theorem's condition is not satisfied. When software rescales an image (the same process that creates the thumbnail shown in the lower image) it, in effect, runs the image through a low-pass filter first and then downsamples the image to result in a smaller image that does not exhibit the moiré pattern. The top image is what happens when the image is downsampled without low-pass filtering: aliasing results.

The application of the sampling theorem to images should be made with care. For example, the sampling process in any standard image sensor (CCD or CMOS camera) is relatively far from the ideal sampling which would measure the image intensity at a single point. Instead these devices have a relatively large sensor area at each sample point in order to obtain sufficient amount of light. In other words, any detector has a finite-width point spread function. The analog optical image intensity function which is sampled by the sensor device is not in general bandlimited, and the non-ideal sampling is itself a useful type of low-pass filter, though not always sufficient to remove enough high frequencies to sufficiently reduce aliasing. When the area of the sampling spot (the size of the pixel sensor) is not large enough to provide sufficient anti-aliasing, a separate anti-aliasing filter (optical low-pass filter) is typically included in a camera system to further blur the optical image. Despite images having these problems in relation to the sampling theorem, the theorem can be used to describe the basics of down and up sampling of images.

Downsampling

When a signal is downsampled, the sampling theorem can be invoked via the artifice of resampling a hypothetical continuous-time reconstruction. The Nyquist criterion must still be satisfied with respect to the new lower sampling frequency in order to avoid aliasing. To meet the requirements of the theorem, the signal must usually pass through a low-pass filter of appropriate cutoff frequency as part of the downsampling operation. This low-pass filter, which prevents aliasing, is called an anti-aliasing filter.

Critical frequency

Fig.7: A family of sinusoids at the critical frequency, all having the same sample sequences of alternating +1 and –1. That is, they all are aliases of each other, even though their frequency is not above half the sample rate.

To illustrate the necessity of fs > 2B, consider the sinusoid:

x(t) = \cos(2 \pi B t + \theta )\ = \ \cos(2 \pi B t)\cos(\theta ) - \sin(2 \pi B t)\sin(\theta ).

With fs = 2B or equivalently T = 1/(2B), the samples are given by:


\begin{align}
x(nT) &= \cos(\pi n)\cos(\theta ) - \underbrace{\sin(\pi n)}_{0}\sin(\theta ) = \cos(\pi n)\cos(\theta ).
\end{align}

Those samples cannot be distinguished from the samples of:

x_A(t) = \cos(2 \pi B t)\cos(\theta ).\,

But for any θ such that sin(θ) ≠ 0, x(t) and xA(t) have different amplitudes and different phase. This and other ambiguities are the reason for the strict inequality of the sampling theorem's condition.

Mathematical basis for the theorem

Fig.8: Spectrum, Xs(f), of a properly sampled bandlimited signal (blue) and images (green) that do not overlap. A "brick-wall" low-pass filter, H(f), removes the images, leaves the original spectrum, X(f), and recovers the original signal from the samples.

From Figures 3 and 8, it is apparent that when there is no overlap of the copies (aka "images") of X(f), the k=0 term of Xs(f) can be recovered by the product:

X(f) = H(f) \cdot X_s(f),\,      where:
H(f) = \begin{cases}1 & |f| < B \\ 0 & |f| > f_s - B. \end{cases}

H(f) need not be precisely defined in the region [B, fs-B], because Xs(f) is zero in that region. However, the worst case is when B = fs/2, the Nyquist frequency.  A function that is sufficient for that and all less severe cases is:

H(f) = \mathrm{rect} \left(\frac{f}{f_s} \right) = \begin{cases}1 & |f| < \frac{f_s}{2} \\ 0 & |f| > \frac{f_s}{2}, \end{cases}

where \mathrm{rect}(u)\ is the rectangular function.

Therefore:

X(f) = \mathrm{rect} \left(\frac{f}{f_s} \right) \cdot X_s(f)\
 = \mathrm{rect} (Tf) \cdot T \sum_{n=-\infty}^{\infty} x(nT)\ e^{-i 2\pi n T f}      (from  Eq.1, above).
 = T \sum_{n=-\infty}^{\infty} x(nT)\cdot \mathrm{rect} (Tf) \cdot e^{-i 2\pi n T f}.

The original function that was sampled can be recovered by an inverse Fourier transform:

x(t) = \mathcal{F}^{-1}\left \{ T \sum_{n=-\infty}^{\infty} x(nT)\cdot \mathrm{rect} (Tf) \cdot e^{-i 2\pi n T f}\right\}
= T \sum_{n=-\infty}^{\infty} x(nT)\cdot \underbrace{\mathcal{F}^{-1}\left \{ \mathrm{rect}(Tf) \cdot e^{-i 2\pi n T f}\right\}}_{\frac{1}{T}\cdot \mathrm{sinc} \left( \frac{t - nT}{T}\right)}     [3]
= \sum_{n=-\infty}^{\infty} x(nT)\cdot \mathrm{sinc} \left( \frac{t - nT}{T}\right),

which is the Whittaker–Shannon interpolation formula. It shows explicitly how the samples, x(nT), can be combined to reconstruct x(t).

Shannon's original proof

The original proof presented by Shannon is elegant and quite brief, but it offers less intuitive insight into the subtleties of aliasing, both unintentional and intentional. Quoting Shannon's original paper, which uses f for the function, F for the spectrum, and W for the bandwidth limit:

Let F(\omega) be the spectrum of f(t). Then
f(t)\, = {1 \over 2\pi} \int_{-\infty}^{\infty} F(\omega) e^{i\omega t}\;d\omega \
= {1 \over 2\pi} \int_{-2\pi W}^{2\pi W} F(\omega) e^{i\omega t}\;d\omega \
since F(\omega) is assumed to be zero outside the band W. If we let
t = {n \over {2W}}
where n is any positive or negative integer, we obtain
f \left({n \over {2W}} \right) = {1 \over 2\pi} \int_{-2\pi W}^{2\pi W} F(\omega) e^{i\omega {n \over {2W}}}\;d\omega.
On the left are values of f(t) at the sampling points. The integral on the right will be recognized as essentially the nth coefficient in a Fourier-series expansion of the function F(\omega), taking the interval –W to W as a fundamental period. This means that the values of the samples f(n/2W) determine the Fourier coefficients in the series expansion of F(\omega). Thus they determine F(\omega), since F(\omega) is zero for frequencies greater than W, and for lower frequencies F(\omega) is determined if its Fourier coefficients are determined. But F(\omega) determines the original function f(t) completely, since a function is determined if its spectrum is known. Therefore the original samples determine the function f(t) completely.

Shannon's proof of the theorem is complete at that point, but he goes on to discuss reconstruction via sinc functions, what we now call the Whittaker–Shannon interpolation formula as discussed above. He does not derive or prove the properties of the sinc function, but these would have been familiar to engineers reading his works at the time, since the Fourier pair relationship between rect (the rectangular function) and sinc was well known. Quoting Shannon:

Let x_n be the nth sample. Then the function f(t) is represented by:
f(t) = \sum_{n=-\infty}^{\infty}x_n{\sin \pi(2Wt-n) \over \pi(2Wt-n)}.

As in the other proof, the existence of the Fourier transform of the original signal is assumed, so the proof does not say whether the sampling theorem extends to bandlimited stationary random processes.

Sampling of non-baseband signals

As discussed by Shannon:[1]

A similar result is true if the band does not start at zero frequency but at some higher value, and can be proved by a linear translation (corresponding physically to single-sideband modulation) of the zero-frequency case. In this case the elementary pulse is obtained from sin(x)/x by single-side-band modulation.

That is, a sufficient no-loss condition for sampling signals that do not have baseband components exists that involves the width of the non-zero frequency interval as opposed to its highest frequency component. See Sampling (signal processing) for more details and examples.

A bandpass condition is that X(f) = 0, for all nonnegative f outside the open band of frequencies:

 \left(\frac{N}2f_\mathrm{s},\frac{N+1}2f_\mathrm{s}\right),

for some nonnegative integer N. This formulation includes the normal baseband condition as the case N=0.

The corresponding interpolation function is the impulse response of an ideal brick-wall bandpass filter (as opposed to the ideal brick-wall lowpass filter used above) with cutoffs at the upper and lower edges of the specified band, which is the difference between a pair of lowpass impulse responses:

(N+1)\,\operatorname{sinc} \left(\frac{(N+1)t}T\right) - N\,\operatorname{sinc}\left( \frac{Nt}T \right).

Other generalizations, for example to signals occupying multiple non-contiguous bands, are possible as well. Even the most generalized form of the sampling theorem does not have a provably true converse. That is, one cannot conclude that information is necessarily lost just because the conditions of the sampling theorem are not satisfied; from an engineering perspective, however, it is generally safe to assume that if the sampling theorem is not satisfied then information will most likely be lost.

Nonuniform sampling

The sampling theory of Shannon can be generalized for the case of nonuniform samples, that is, samples not taken equally spaced in time. Shannon sampling theory for non-uniform sampling states that a band-limited signal can be perfectly reconstructed from its samples if the average sampling rate satisfies the Nyquist condition[4]. Therefore, although uniformly spaced samples may result in easier reconstruction algorithms, it is not a necessary condition for perfect reconstruction.

Beyond Nyquist

The Nyquist–Shannon sampling theorem is also known to be a sufficient condition (i.e., it is not necessary). The field of compressed sensing provides a stricter sampling condition when the underlying signal is known to be sparse. Compressed sensing specifically yields a sub-Nyquist sampling criterion. The knowledge of the sparsity or compressibility of the signal provides some bounds on the number of samples needed to allow for reconstruction of the signal. In compressed sensing, the reconstruction of the original signal from the samples involves non-linear solvers.

Historical background

The sampling theorem was implied by the work of Harry Nyquist in 1928 ("Certain topics in telegraph transmission theory"), in which he showed that up to 2B independent pulse samples could be sent through a system of bandwidth B; but he did not explicitly consider the problem of sampling and reconstruction of continuous signals. About the same time, Karl Küpfmüller showed a similar result,[5] and discussed the sinc-function impulse response of a band-limiting filter, via its integral, the step response Integralsinus; this bandlimiting and reconstruction filter that is so central to the sampling theorem is sometimes referred to as a Küpfmüller filter (but seldom so in English).

The sampling theorem, essentially a dual of Nyquist's result, was proved by Claude E. Shannon in 1949 ("Communication in the presence of noise"). V. A. Kotelnikov published similar results in 1933 ("On the transmission capacity of the 'ether' and of cables in electrical communications", translation from the Russian), as did the mathematician E. T. Whittaker in 1915 ("Expansions of the Interpolation-Theory", "Theorie der Kardinalfunktionen"), J. M. Whittaker in 1935 ("Interpolatory function theory"), and Gabor in 1946 ("Theory of communication").

Other discoverers

Others who have independently discovered or played roles in the development of the sampling theorem have been discussed in several historical articles, for example by Jerri[6] and by Lüke.[7] For example, Lüke points out that H. Raabe, an assistant to Küpfmüller, proved the theorem in his 1939 Ph.D. dissertation; the term Raabe condition came to be associated with the criterion for unambiguous representation (sampling rate greater than twice the bandwidth).

Meijering[8] mentions several other discoverers and names in a paragraph and pair of footnotes:

As pointed out by Higgins [135], the sampling theorem should really be considered in two parts, as done above: the first stating the fact that a bandlimited function is completely determined by its samples, the second describing how to reconstruct the function using its samples. Both parts of the sampling theorem were given in a somewhat different form by J. M. Whittaker [350, 351, 353] and before him also by Ogura [241, 242]. They were probably not aware of the fact that the first part of the theorem had been stated as early as 1897 by Borel [25].27 As we have seen, Borel also used around that time what became known as the cardinal series. However, he appears not to have made the link [135]. In later years it became known that the sampling theorem had been presented before Shannon to the Russian communication community by Kotel'nikov [173]. In more implicit, verbal form, it had also been described in the German literature by Raabe [257]. Several authors [33, 205] have mentioned that Someya [296] introduced the theorem in the Japanese literature parallel to Shannon. In the English literature, Weston [347] introduced it independently of Shannon around the same time.28

27 Several authors, following Black [16], have claimed that this first part of the sampling theorem was stated even earlier by Cauchy, in a paper [41] published in 1841. However, the paper of Cauchy does not contain such a statement, as has been pointed out by Higgins [135].

28 As a consequence of the discovery of the several independent introductions of the sampling theorem, people started to refer to the theorem by including the names of the aforementioned authors, resulting in such catchphrases as “the Whittaker-Kotel’nikov-Shannon (WKS) sampling theorem" [155] or even "the Whittaker-Kotel'nikov-Raabe-Shannon-Someya sampling theorem" [33]. To avoid confusion, perhaps the best thing to do is to refer to it as the sampling theorem, "rather than trying to find a title that does justice to all claimants" [136].

Why Nyquist?

Exactly how, when, or why Harry Nyquist had his name attached to the sampling theorem remains obscure. The term Nyquist Sampling Theorem (capitalized thus) appeared as early as 1959 in a book from his former employer, Bell Labs,[9] and appeared again in 1963,[10] and not capitalized in 1965.[11] It had been called the Shannon Sampling Theorem as early as 1954,[12] but also just the sampling theorem by several other books in the early 1950s.

In 1958, Blackman and Tukey[13] cited Nyquist's 1928 paper as a reference for the sampling theorem of information theory, even though that paper does not treat sampling and reconstruction of continuous signals as others did. Their glossary of terms includes these entries:

Sampling theorem (of information theory)
Nyquist's result that equi-spaced data, with two or more points per cycle of highest frequency, allows reconstruction of band-limited functions. (See Cardinal theorem.)
Cardinal theorem (of interpolation theory)
A precise statement of the conditions under which values given at a doubly infinite set of equally spaced points can be interpolated to yield a continuous band-limited function with the aid of the function
\frac{\sin (x - x_i)}{x - x_i}.

Exactly what "Nyquist's result" they are referring to remains mysterious.

When Shannon stated and proved the sampling theorem in his 1949 paper, according to Meijering[8] "he referred to the critical sampling interval T = 1/(2W) as the Nyquist interval corresponding to the band W, in recognition of Nyquist’s discovery of the fundamental importance of this interval in connection with telegraphy." This explains Nyquist's name on the critical interval, but not on the theorem.

Similarly, Nyquist's name was attached to Nyquist rate in 1953 by Harold S. Black:[14]

"If the essential frequency range is limited to B cycles per second, 2B was given by Nyquist as the maximum number of code elements per second that could be unambiguously resolved, assuming the peak interference is less half a quantum step. This rate is generally referred to as signaling at the Nyquist rate and 1/(2B) has been termed a Nyquist interval." (bold added for emphasis; italics as in the original)

According to the OED, this may be the origin of the term Nyquist rate. In Black's usage, it is not a sampling rate, but a signaling rate.

See also

Notes

  1. 1.0 1.1 C. E. Shannon, "Communication in the presence of noise", Proc. Institute of Radio Engineers, vol. 37, no. 1, pp. 10–21, Jan. 1949. Reprint as classic paper in: Proc. IEEE, vol. 86, no. 2, (Feb. 1998)
  2. supp(X) means:   the support of X
  3. The time-domain form follows from rows 202 and 102 of the transform tables
  4. Nonuniform Sampling, Theory and Practice (ed. F. Marvasti), Kluwer Academic/Plenum Publishers, New York, 2000
  5. K. Küpfmüller, "Über die Dynamik der selbsttätigen Verstärkungsregler", Elektrische Nachrichtentechnik, vol. 5, no. 11, pp. 459-467, 1928. (German)
      K. Küpfmüller, On the dynamics of automatic gain controllers, Elektrische Nachrichtentechnik, vol. 5, no. 11, pp. 459-467, 1928 (English translation 2005).
  6. Abdul Jerri, The Shannon Sampling Theorem—Its Various Extensions and Applications: A Tutorial Review, Proceedings of the IEEE, 65:1565–1595, Nov. 1977. See also Correction to "The Shannon sampling theorem—Its various extensions and applications: A tutorial review", Proceedings of the IEEE, 67:695, April 1979
  7. Hans Dieter Lüke, The Origins of the Sampling Theorem, IEEE Communications Magazine, pp.106–108, April 1999.
  8. 8.0 8.1 Erik Meijering, A Chronology of Interpolation From Ancient Astronomy to Modern Signal and Image Processing, Proc. IEEE, 90, 2002.
  9. Members of the Technical Staff of Bell Telephone Lababoratories (1959). Transmission Systems for Communications. AT&T. pp. 26–4 (Vol.2). 
  10. Ernst Adolph Guillemin (1963). Theory of Linear Physical Systems. Wiley. http://books.google.com/books?id=jtI-AAAAIAAJ&q=Nyquist-sampling-theorem+date:0-1965&dq=Nyquist-sampling-theorem+date:0-1965&as_brr=0&ei=ief5Ru-DJqegowLyiJTpBw&pgis=1. 
  11. Richard A. Roberts and Ben F. Barton, Theory of Signal Detectability: Composite Deferred Decision Theory, 1965.
  12. Truman S. Gray, Applied Electronics: A First Course in Electronics, Electron Tubes, and Associated Circuits, 1954.
  13. R. B. Blackman and J. W. Tukey, The Measurement of Power Spectra : From the Point of View of Communications Engineering, New York: Dover, 1958.
  14. Harold S. Black, Modulation Theory, 1953

References

External links